Goto

Collaborating Authors

 North West Shelf


Multimodal Generative Flows for LHC Jets

Faroughy, Darius A., Opper, Manfred, Ojeda, Cesar

arXiv.org Artificial Intelligence

Generative modeling of high-energy collisions at the Large Hadron Collider (LHC) offers a data-driven route to simulations, anomaly detection, among other applications. A central challenge lies in the hybrid nature of particle-cloud data: each particle carries continuous kinematic features and discrete quantum numbers such as charge and flavor. We introduce a transformer-based multimodal flow that extends flow-matching with a continuous-time Markov jump bridge to jointly model LHC jets with both modalities. Trained on CMS Open Data, our model can generate high fidelity jets with realistic kinematics, jet substructure and flavor composition.


IONext: Unlocking the Next Era of Inertial Odometry

Zhang, Shanshan, Zhang, Qi, Wang, Siyue, Wen, Tianshui, Wu, Liqin, Zhou, Ziheng, Hong, Xuemin, Peng, Ao, Zheng, Lingxiang, Yang, Yu

arXiv.org Artificial Intelligence

Researchers have increasingly adopted Transformer-based models for inertial odometry. While Transformers excel at modeling long-range dependencies, their limited sensitivity to local, fine-grained motion variations and lack of inherent inductive biases often hinder localization accuracy and generalization. Recent studies have shown that incorporating large-kernel convolutions and Transformer-inspired architectural designs into CNN can effectively expand the receptive field, thereby improving global motion perception. Motivated by these insights, we propose a novel CNN-based module called the Dual-wing Adaptive Dynamic Mixer (DADM), which adaptively captures both global motion patterns and local, fine-grained motion features from dynamic inputs. This module dynamically generates selective weights based on the input, enabling efficient multi-scale feature aggregation. To further improve temporal modeling, we introduce the Spatio-Temporal Gating Unit (STGU), which selectively extracts representative and task-relevant motion features in the temporal domain. This unit addresses the limitations of temporal modeling observed in existing CNN approaches. Built upon DADM and STGU, we present a new CNN-based inertial odometry backbone, named Next Era of Inertial Odometry (IONext). Extensive experiments on six public datasets demonstrate that IONext consistently outperforms state-of-the-art (SOTA) Transformer- and CNN-based methods. For instance, on the RNIN dataset, IONext reduces the average ATE by 10% and the average RTE by 12% compared to the representative model iMOT.


Tethered Multi-Robot Systems in Marine Environments

Buchholz, Markus, Carlucho, Ignacio, Grimaldi, Michele, Petillot, Yvan R.

arXiv.org Artificial Intelligence

This paper introduces a novel simulation framework for evaluating motion control in tethered multi-robot systems within dynamic marine environments. Specifically, it focuses on the coordinated operation of an Autonomous Underwater Vehicle (AUV) and an Autonomous Surface Vehicle(ASV). The framework leverages GazeboSim, enhanced with realistic marine environment plugins and ArduPilots SoftwareIn-The-Loop (SITL) mode, to provide a high-fidelity simulation platform. A detailed tether model, combining catenary equations and physical simulation, is integrated to accurately represent the dynamic interactions between the vehicles and the environment. This setup facilitates the development and testing of advanced control strategies under realistic conditions, demonstrating the frameworks capability to analyze complex tether interactions and their impact on system performance.


Rethinking Machine Unlearning in Image Generation Models

Liu, Renyang, Feng, Wenjie, Zhang, Tianwei, Zhou, Wei, Cheng, Xueqi, Ng, See-Kiong

arXiv.org Artificial Intelligence

With the surge and widespread application of image generation models, data privacy and content safety have become major concerns and attracted great attention from users, service providers, and policymakers. Machine unlearning (MU) is recognized as a cost-effective and promising means to address these challenges. Despite some advancements, image generation model unlearning (IGMU) still faces remarkable gaps in practice, e.g., unclear task discrimination and unlearning guidelines, lack of an effective evaluation framework, and unreliable evaluation metrics. These can hinder the understanding of unlearning mechanisms and the design of practical unlearning algorithms. We perform exhaustive assessments over existing state-of-the-art unlearning algorithms and evaluation standards, and discover several critical flaws and challenges in IGMU tasks. Driven by these limitations, we make several core contributions, to facilitate the comprehensive understanding, standardized categorization, and reliable evaluation of IGMU. Specifically, (1) We design CatIGMU, a novel hierarchical task categorization framework. It provides detailed implementation guidance for IGMU, assisting in the design of unlearning algorithms and the construction of testbeds. (2) We introduce EvalIGMU, a comprehensive evaluation framework. It includes reliable quantitative metrics across five critical aspects. (3) We construct DataIGM, a high-quality unlearning dataset, which can be used for extensive evaluations of IGMU, training content detectors for judgment, and benchmarking the state-of-the-art unlearning algorithms. With EvalIGMU and DataIGM, we discover that most existing IGMU algorithms cannot handle the unlearning well across different evaluation dimensions, especially for preservation and robustness. Code and models are available at https://github.com/ryliu68/IGMU.


FamilyTool: A Multi-hop Personalized Tool Use Benchmark

Wang, Yuxin, Guo, Yiran, Zheng, Yining, Yin, Zhangyue, Chen, Shuo, Yang, Jie, Chen, Jiajun, Li, Yuan, Huang, Xuanjing, Qiu, Xipeng

arXiv.org Artificial Intelligence

The integration of tool learning with Large Language Models (LLMs) has expanded their capabilities in handling complex tasks by leveraging external tools. However, existing benchmarks for tool learning inadequately address critical real-world personalized scenarios, particularly those requiring multi-hop reasoning and inductive knowledge adaptation in dynamic environments. To bridge this gap, we introduce FamilyTool, a novel benchmark grounded in a family-based knowledge graph (KG) that simulates personalized, multi-hop tool use scenarios. FamilyTool, including base and extended datasets, challenges LLMs with queries spanning from 1 to 4 relational hops (e.g., inferring familial connections and preferences) and 2 to 6 hops respectively, and incorporates an inductive KG setting where models must adapt to unseen user preferences and relationships without re-training, a common limitation in prior approaches that compromises generalization. We further propose KGETool: a simple KG-augmented evaluation pipeline to systematically assess LLMs' tool use ability in these settings. Experiments reveal significant performance gaps in state-of-the-art LLMs, with accuracy dropping sharply as hop complexity increases and inductive scenarios exposing severe generalization deficits. These findings underscore the limitations of current LLMs in handling personalized, evolving real-world contexts and highlight the urgent need for advancements in tool-learning frameworks. FamilyTool serves as a critical resource for evaluating and advancing LLM agents' reasoning, adaptability, and scalability in complex, dynamic environments. Code and dataset are available at \href{https://github.com/yxzwang/FamilyTool}{https://github.com/yxzwang/FamilyTool}.


Information Theory-Guided Heuristic Progressive Multi-View Coding

Li, Jiangmeng, Gao, Hang, Qiang, Wenwen, Zheng, Changwen

arXiv.org Artificial Intelligence

Multi-view representation learning aims to capture comprehensive information from multiple views of a shared context. Recent works intuitively apply contrastive learning to different views in a pairwise manner, which is still scalable: view-specific noise is not filtered in learning view-shared representations; the fake negative pairs, where the negative terms are actually within the same class as the positive, and the real negative pairs are coequally treated; evenly measuring the similarities between terms might interfere with optimization. Importantly, few works study the theoretical framework of generalized self-supervised multi-view learning, especially for more than two views. To this end, we rethink the existing multi-view learning paradigm from the perspective of information theory and then propose a novel information theoretical framework for generalized multi-view learning. Guided by it, we build a multi-view coding method with a three-tier progressive architecture, namely Information theory-guided hierarchical Progressive Multi-view Coding (IPMC). In the distribution-tier, IPMC aligns the distribution between views to reduce view-specific noise. In the set-tier, IPMC constructs self-adjusted contrasting pools, which are adaptively modified by a view filter. Lastly, in the instance-tier, we adopt a designed unified loss to learn representations and reduce the gradient interference. Theoretically and empirically, we demonstrate the superiority of IPMC over state-of-the-art methods.


Elucidating STEM Concepts through Generative AI: A Multi-modal Exploration of Analogical Reasoning

Cao, Chen, Ding, Zijian, Lee, Gyeong-Geon, Jiao, Jiajun, Lin, Jionghao, Zhai, Xiaoming

arXiv.org Artificial Intelligence

This study explores the integration of generative artificial intelligence (AI), specifically large language models, with multi-modal analogical reasoning as an innovative approach to enhance science, technology, engineering, and mathematics (STEM) education. We have developed a novel system that utilizes the capacities of generative AI to transform intricate principles in mathematics, physics, and programming into comprehensible metaphors. To further augment the educational experience, these metaphors are subsequently converted into visual form. Our study aims to enhance the learners' understanding of STEM concepts and their learning engagement by using the visual metaphors. We examine the efficacy of our system via a randomized A/B/C test, assessing learning gains and motivation shifts among the learners. Our study demonstrates the potential of applying large language models to educational practice on STEM subjects. The results will shed light on the design of educational system in terms of harnessing AI's potential to empower educational stakeholders.


A Brief History of AI, From French Philosophy to Self-Driving Cars - Dell Technologies

#artificialintelligence

And yet, AI's current automated task-mastering was first posited by the French philosopher René Descartes almost 400 years ago. Descartes, who famously coined, "I think, therefore I am," pondered about the ability of machines to reason. While machines may be able to "do some things as well, or better, than humans, they would inevitably fail in others," whereas human reason can universally adapt to any task. Though Descartes' idea of machines differs from today's reality, some say he threw down the gauntlet for what we now refer to as general AI--or machines that can think like humans. Though Descartes' idea of machines differs from today's reality, some say he threw down the gauntlet for what we now refer to as general AI--or machines that can think like humans.


Would you ride Lyft in a self-driving car? Waymo's autonomous minivans are breaking in to ride-sharing

USATODAY - Tech Top Stories

Waymo is launching its first self-driving car service in Phoenix Arizona called Waymo One. Waymo, a subsidiary of Google-parent Alphabet which is developing autonomous vehicles and related services, has officially expanded its reach and is now making some of its self-driving minivans available for customers of ride-share operator Lyft. The rides are restricted to a small area outside Phoenix, where Waymo has been testing self-driving vehicles and has started its own autonomous ride-share service, Waymo One. Waymo's limited partnership with Lyft is the latest example of the company branching out to work with more companies as it develops autonomous vehicles and services. Earlier this month, Waymo struck a deal with Nissan and Renault to build self-driving vehicles for those automakers.


3 Top Artificial Intelligence Stocks to Watch in June

#artificialintelligence

For technology investors, artificial intelligence (AI) is the next frontier. And for good reason, as Accenture finds investments in AI will turbocharge the economy, boosting productivity in the U.S. by more than a third and nearly doubling GDP growth rates. Early-stage AI investors are in a key position for multidecade returns. We asked a trio of our Motley Fool contributors to highlight three companies well poised to take advantage of AI's growth. Anders Bylund (IBM): With $78.7 billion of trailing revenues and $17.7 billion in EBITDA profits, IBM is an instant giant in pretty much any niche it decides to address.